Robust multi-class boosting
نویسنده
چکیده
Boosting approaches are based on the idea that high-quality learning algorithms can be formed by repeated use of a “weak-learner”, which is required to perform only slightly better than random guessing. It is known that Boosting can lead to drastic improvements compared to the individual weak-learner. For two-class problems it has been shown that the original Boosting algorithm, called AdaBoost, is quite unaffected by overfitting. However, for the case of noisy data, it is also understood that AdaBoost can be improved considerably by introducing some regularization technique. In speech-related problems one often considers multi-class problems and Boosting formulations have been used successfully to solve them. I review existing multi-class boosting algorithms, which have been much less analyzed and explored than the twoclass pendants. In this work I extend these methods to derive new boosting algorithms which are more robust against outliers and noise in the data and are able to exploit prior knowledge about relationships between the classes.
منابع مشابه
Robust Multi-View Boosting with Priors
Many learning tasks for computer vision problems can be described by multiple views or multiple features. These views can be exploited in order to learn from unlabeled data, a.k.a. “multi-view learning”. In these methods, usually the classifiers iteratively label each other a subset of the unlabeled data and ignore the rest. In this work, we propose a new multi-view boosting algorithm that, unl...
متن کاملAn Empirical Evaluation of Four Algorithms for Multi-Class Classification: Mart, ABC-Mart, Robust LogitBoost, and ABC-LogitBoost
This empirical study is mainly devoted to comparing four tree-based boosting algorithms: mart, abc-mart, robust logitboost, and abc-logitboost, for multi-class classification on a variety of publicly available datasets. Some of those datasets have been thoroughly tested in prior studies using a broad range of classification algorithms including SVM, neural nets, and deep learning. In terms of t...
متن کاملRobust LogitBoost and Adaptive Base Class (ABC) LogitBoost
Logitboost is an influential boosting algorithm for classification. In this paper, we develop robust logitboost to provide an explicit formulation of tree-split criterion for building weak learners (regression trees) for logitboost. This formulation leads to a numerically stable implementation of logitboost. We then propose abc-logitboost for multi-class classification, by combining robust logi...
متن کاملA Direct Approach to Multi-class Boosting and Extensions
Boosting methods combine a set of moderately accurate weak learners to form a highly accurate predictor. Despite the practical importance of multi-class boosting, it has received far less attention than its binary counterpart. In this work, we propose a fully-corrective multi-class boosting formulation which directly solves the multi-class problem without dividing it into multiple binary classi...
متن کاملAppendix: Sharing Features in Multi-class Boosting via Group Sparsity
In this document we provide a complete derivation for multi-class boosting with group sparsity and a full explanation of admm algorithm presented in the main paper. 1 Multi-class boosting with group sparsity We first provide the derivation for multi-class logistic loss with 1,2-norm. We then show the difference between our boosting with 1,2-norm and 1,∞-norm. We then briefly discuss our group s...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2003